The right strategy wins the war WeatherShop.com Gifts, gadgets, weather stations, software and more...click here!\
The Blogosphere
Tuesday, June 02, 2009
How Many IPCC (and Other) Scientists Fabricate and Falsify Research?

I am at the Third International Conference on Climate Change in DC with talks and meetings scheduled for tomorrow. I will post on results midweek. Meanwhile see this note to Dr. Benny Peiser of CCNet by Norm Kalmanovitch.

By Norm Kalmanovitch

The measured global temperature record which started around 1856 shows that the Earth was in a warming cycle until around 1880. The CO2 record shows that CO2 was increasing by about 0.21ppmv/year over this period. During the cooling cycle which followed from 1880 to 1910, the CO2 concentration increased at a rate of about 0.30ppmv/year.

The next warming cycle from 1910 to 1942 saw a dramatic increase in global temperature, but the rate of increase in CO2 concentration only grew to 0.33ppmv over this time period. The well documented global cooling period from 1942 to 1975 that had the world concerned about an impending return to the equivalent of the Little Ice Age, had a contemporaneous rise in atmospheric CO2 that equated to 0.63ppmv/year; almost twice the increase in CO2 of the precious warming cycle.

During the warming that took place from 1975 to 1998, the rate of CO2 increase took another dramatic jump to 1.54ppmv/year, but this was followed by an increase to 1.91ppmv/year that we are currently experiencing during the present ongoing cooling cycle. Each successive cooling cycle has had an increase in the rate of CO2 growth over the previous warming cycle, indicating that there is no possible correlation of CO2 with global warming.

In 1988 Hansen et al published a paper “Global Climate Changes as Forecast by Goddard Institute for Space Studies Three-Dimensional Model” in the Journal of Geophysical Research that introduced a “CO2 forcing parameter”. This parameter had no actual physical basis, but was merely based on the assumption that a 100ppmv increase in CO2 was directly and primarily responsible for the measured increase in global temperature of 0.6C that had been observed over the past century.

This assumption ignored the fact that over this time period there was both cooling and warming concurrent with rising CO2 concentration, and considering that this paper was published just 13 years after a 33 year cooling trend that also had a concurrent increase in CO2 concentration there is no possible valid rational for this assumption. Essentially in the 46year period from 1942 to when the paper was published in 1988, there were 33 years of cooling and only 13 years of warming concurrent with increases in CO2, yet the models used a forcing parameter that directly related only warming to CO2 concentration increases.

With no basis in fact, this parameter is entirely a fabrication, and the projections of climate models that are based on this fabricated parameter are also meaningless fabrications. In addition to the fabrication, there is a bit of scientific fraud in the creation of this CO2 forcing parameter.

The Earth had been warming since the Little Ice Age at a rate of about 0.5C/century. The temperature value that went into determining the CO2 forcing parameter was 0.6C, with the difference from the 0.5C/century value likely due to the urban heat island effect. Even if this difference was directly due to CO2 increases, the difference between the observed temperature and the natural warming since the Little Ice Age is only 0.1C but the full 0.6C was used to fabricate the forcing parameter.

It seems that one fabrication leads to another, and when it became obvious that the natural warming of 0.5C/century since the Little Ice Age demonstrated the obvious deficiency in this forcing parameter of the climate models, the MBH98 temperature proxy also known as the “hockey stick” was fabricated to remove the Little Ice Age and allow the full 0.6C temperature increase to be related to CO2 increases. Considering that the climate models are the only support for the AGW premise, and the AGW premise is the only support for the climate models, exposing this simple fabrication is all that needs to be done to put an end to this circular argument that forms the basis for the entire climate change lunacy.

Posted on 06/02 at 02:45 AM
(357) TrackbacksPermalink


New Paper “Arctic Air Temperature Change Amplification”

By Roger Pielke Sr. Climate Science

An excellent new paper is “in press” for Geophysical Research Letters (GRL) which documents the major role of regional atmospheric/ocean circulation pattern changes on regional multi-decadal climate variability (e.g. see What is the Importance to Climate of Heterogeneous Spatial Trends in Tropospheric Temperatures?).

This paper supports the finding that long term variations in atmospheric/ocean circulations (such as the Atlantic Multidecadal Oscillation, the Pacific Decadal Oscillation, ENSO, etc) cause regional changes in temperatures over this time period, and that these changes have a significant natural cause. Such a perspective supports the views of Joe D’Aleo (see) and Bill Gray (see) and Roy Spencer (e.g. see).

The paper is: Chylek Petr, Chris K. Folland, Glen Lesins, Manvendra K. Dubeys, and Muyin Wang: 2009: “Arctic air temperature change amplification and the Atlantic Multidecadal Oscillation”. Geophysical Research Letters (in press).

The abstract reads

“Understanding Arctic temperature variability is essential for assessing possible future melting of the Greenland ice sheet, Arctic sea ice and Arctic permafrost. Temperature trend reversals in 1940 and 1970 separate two Arctic warming periods (1910-1940 and 1970-2008) by a significant 1940-1970 cooling period. Analyzing temperature records of the Arctic meteorological stations we find that (a) the Arctic amplification (ratio of the Arctic to global temperature trends) is not a constant but varies in time on a multi-decadal time scale, (b) the Arctic warming from 1910-1940 proceeded at a significantly faster rate than the current 1970-2008 warming, and (c) the Arctic temperature changes are highly correlated with the Atlantic Multi-decadal Oscillation (AMO) suggesting the Atlantic Ocean thermohaline circulation is linked to the Arctic temperature variability on a multi decadal time scale.”

Text in this paper includes: “In the following analysis we confirm that the Arctic has indeed warmed during the 1970-2008 period by a factor of two to three faster than the global mean in agreement with model predictions but the reasons may not be entirely anthropogenic. We find that the ratio of the Arctic to global temperature change was much larger during the years 1910-1970.” “We consequently propose that the AMO is a major factor affecting inter-decadal variations of Arctic temperature and explaining [the] high value of the Arctic to global temperature trend ratio during the cooling period of 1940-1970.”

“Our analysis suggests that the ratio of the Arctic to global temperature change varies on [a] multi-decadal time scale. The commonly held assumption of a factor of 2-3 for the Arctic amplification has been valid only for the current warming period 1970-2008. The Arctic region did warm considerably faster during the 1910-1940 warming compared to the current 1970-2008 warming rate. During the cooling from 1940-1970 the Arctic amplification was extremely high, between 9 and 13. The Atlantic Ocean thermohaline circulation multi-decadal variability is suggested as a major cause of Arctic temperature variation. Further analyses of long coupled model runs will be critical to resolve the influence of the ocean thermohaline circulation and other natural climate variations on Arctic climate and to determine whether natural climate variability will make the Arctic more or less vulnerable to anthropogenic global warming.” See Roger’s full post here.

Icecap Note: What we found:

image
Larger image here.

Posted on 06/02 at 02:09 AM
(109) TrackbacksPermalink


Friday, May 29, 2009
Purists Who Demand that Facts be “Correct”

By Roger Pielke, Jr., Prometheus

Over at Seed Magazine in a collection of views on “framing,” Penn State climatologist Michael Mann explains why it was necessary to misrepresent what the IPCC does on the cover of his co-authored book titled “Dire Predictions: Understanding Global Warming”:

Often, in our communication efforts, scientists are confronted with critical issues of language and framing. A case in point is a book I recently co-authored with Penn State colleague Lee Kump, called Dire Predictions: Understanding Global Warming. The purists among my colleagues would rightly point out that the potential future climate changes we describe, are, technically speaking, projections rather than predictions because the climate models are driven by hypothetical pathways of future fossil fuel burning (i.e. conceivable but not predicted futures). But Dire Projections doesn’t quite roll off the tongue. And it doesn’t convey; - in the common vernacular; - what the models indicate: Climate change could pose a very real threat to society and the environment. In this case, use of the more technically “correct” term is actually less likely to convey the key implications to a lay audience.

image

As one of those “purists” who would like to receive information that is technically “correct” I probably can read that book by its cover.

In contrast, in another commentary on framing at Seed, ASU science policy expert Clark Miller suggests an alternative, richer view of framing: Two competing models of framing exist. The first views framing as a tactical choice in communication. Spinning information to comport with culturally embedded narratives purportedly raises its credibility with target audiences. This model presumes an ignorant and uninformed public, with all the dangers that implies for democracy. I reject this model.

The second model views framing, instead, as how humans make sense of and give meaning to events in the world  -  the lens through which they interpret disparate observations, models, data, and evidence in light of their values. This model posits framing as an ineradicable element of reasoning, even in science, and a facility for rich, nuanced storytelling as a foundation for human community.

Both models recognize that humans structure their understanding of policy through narrative and story. Rather than exploiting this structure for political gain, however, the second model acknowledges that any specific policy frame is, at best, partial and incomplete. Any frame reflects only one way of looking at a policy problem, leaving out potentially critical pieces of knowledge and significance.

See post at Prometheus.

Posted on 05/29 at 05:54 AM
(445) TrackbacksPermalink


Thursday, May 28, 2009
Canada Has a Frigid May after a Cold Winter

By Joseph D’Aleo

May has been frigid slowing the planting and emergence of the summer crops in Canada. Late freezes and even snows are still occurring regularly and can be expected the rest of the month.

image
See larger image here.

The chart above shows the May 2009 temperature anomaly through May 24th. Parts of central Canada (areas around Churchill, Manitoba) are running 12-16 degrees F below normal for the month through the 26th (map ends 24th). Every day this month has seen lows below freezing in Churchill and only 6 out of the first 26 days days had highs edge above freezing. The forecast the rest of the month is for more cold with even some snow today in Churchill and again this weekend perhaps further south.

Hudson Bay remains mostly frozen though most of the seasonal melting occurs in June and July most years.

Parts of the south central region were also cold in April averaging 3-5 F below normal. The winter (December to March) was a cold one for southwest and central Canada but warmer in the far northeast.

image
See larger image here.

Meanwhile the arctic ice remains higher this data for any year this decade in a virtual tie with 2004.

image
See larger image here.

Given the polar stratospheric aerosols from Mt Redoubt, and a colder Atlantic and a continued cold Pacific, the recovery from the minimum of 2007 should continue this season.

The global data bases have large gaps in Canada, Africa, South America. So they will not reflect this in their global May anomalies as well as the satellites that see the entire surface - land and ocean excluding high latitude polar.

See pdf here. H/T Climate Depot and Andy for the heads up.

Posted on 05/28 at 12:20 AM
(210) TrackbacksPermalink


Wednesday, May 27, 2009
The MIT Global Warming Gamble

By Dr. Roy Spencer

Climate science took another step backward last week as a new study from the Massachusetts Institute of Technology was announced which claims global warming by 2100 will probably be twice as bad as the United Nations Intergovernmental Panel on Climate Change (IPCC) has predicted.  The research team examined a range of possible climate scenarios which combined various estimates of the sensitivity of the climate system with a range of possible policy decisions to reduce greenhouse gas emissions which (presumably) cause global warming. Without policy action, the group’s model runs “indicate a median probability of surface warming of 5.2 degrees Celsius by 2100, with a 90% probability range of 3.5 to 7.4 degrees”.

Since that average rate of warming (about 0.5 deg. C per decade) is at least 2 times the observed rate of global-average surface temperature rise over the last 30 years, this would require our current spate of no warming to change into very dramatic and sustained warming in the near future. And the longer Mother Nature waits to comply with the MIT group’s demands, the more severe the warming will have to be to meet their projections.

Of course, as readers of this web site will know, the MIT results are totally dependent upon the climate sensitivity that was assumed in the climate model runs that formed the basis for their calculations. And climate modelers can get just about any level of warming they want by simply making a small change in the processes controlling climate sensitivity - especially cloud feedbacks - in those models.

So, since the sensitivity of the climate system is uncertain, these researchers followed the IPCC’s lead of using ‘statistical probability’ as a way of treating that uncertainty. But as I have mentioned before, the use of statistical probabilities in this context is inappropriate. There is a certain climate sensitivity that exists in the real climate system, and it is true that we do not know exactly what that sensitivity is. But this does not mean that our uncertainty over its sensitivity can be translated into some sort of statistical probability.

The use of statistical probabilities by the IPCC and the MIT group does two misleading things: (1) it implies scientific precision where none exists, and (2) it implies the climate system’s response to any change is a “roll of the dice”. We know what the probability of rolling a pair of sixes with dice is, since it is a random event which, when repeated a sufficient number of times, will reveal that probability (1 in 36). But in contrast to this simple example, there is instead a particular climate sensitivity that exists out there in the real climate system. The endless fascination with playing computer games to figure out that climate sensitivity, in my opinion, ends up wasting a lot of time and money.

True, there are many scientists who really do think our tinkering with the climate system through our greenhouse gas emissions is like playing Russian roulette. But the climate system tinkers with itself all the time, and the climate has managed to remain stable. There are indeed internal, chaotic fluctuations in the climate system that might appear to be random, but their effect on the whole climate system are constrained to operate within a certain range. If the climate system really was that sensitive, it would have forced itself into oblivion long ago. The MIT research group pays lip service to relying on “peer-reviewed science”, but it looks like they treat peer-reviewed scientific publications as random events, too. If 99 papers have been published which claim the climate system is VERY sensitive, but only 1 paper has been published that says the climate system is NOT very sensitive, is there then a 99-in-100 (99%) chance that the climate system is very sensitive? NO. As has happened repeatedly in all scientific disciplines, it is often a single research paper that ends up overturning what scientists thought they knew about something.

In climate research, those 99 papers typically will all make the same assumptions, which then pretty much guarantees they will end up arriving at the same conclusions. So, those 99 papers do not constitute independent pieces of evidence. Instead, they might be better described as evidence that ‘group think’ still exists. It turns out that the belief in a sensitive climate is not because of the observational evidence, but in spite of it. You can start to learn more about the evidence for low climate sensitivity (negative feedbacks) here.

image

As the slightly-retouched photo of the MIT research group shown above suggests, I predict that it is only a matter of time before the climate community placing all its bets on the climate models is revealed to be a very bad gamble. See full blog here.

Posted on 05/27 at 02:23 AM
(38) TrackbacksPermalink


Tuesday, May 26, 2009
Tornado Season So Far Not as Bad as 2008

By Joseph D’Aleo on Intellicast

After another La Nina season with again a lot of snow and precipitation in the north central, another relatively active tornado season was expected and so far it has delivered on that promise. However given the La Nina was not as strong and the rebound in the Pacific towards El Nino is a month earlier than last year, the number of storms so far, have been less. It looks like May will fall well short of last May’s 461 tornadoes.

image
Cedar Hill, Texas, Photo credit Pat Skinner, TTU

The annual summary to date can be found here. The tornadoes so far in 2009 have been in the southeast quadrant of the nation. Climatologically, that is where the season normally begins.

image
See larger image here.

In 2008, the tornadoes when all was said and done, were found in all but 4 of the lower 48 states.
image
See larger image here.

As we move into summer, expect the activity to shift north with the jet stream. The march of the season - climatology of tornadoes normally follows this depiction (source here). 

We are below the 5 year average for tornadoes for the season to date, below all but 2005.

image
See larger image here.

The activity was as usual concentrated in what is called tornado alley in the plains, Midwest to the Gulf. You can see in 2008 that the daily events peaked in May with the biggest day on the 23rd of May before falling off in summer as El Nino like conditions developed in the eastern Pacific. That is occurring a month earlier this year and perhaps, that explains the quieter May.

image
See larger image here.

The activity was as usual concentrated in what is called tornado alley in the plains, Midwest to the Gulf.

image
See larger image here.

The reason that this region is most vulnerable is that this is where the combination of moisture from the Gulf of Mexico, dry air in mid levels from Mexico, a strong jet with cold air aloft coming out of the Rockies and a boundary between still cool air to the north and the warm humid summer like air in the south all come together.  Read more and see some useful links here.

By the way, last year, the alarmist media blamed the increase in tornados on global warming. Well guess what this year, stormchasers across Tornado Alley have been frustrated this spring by what seems to be a lack of tornadoes and severe weather.  Indeed, VORTEX2, the largest tornado field study ever, has been running for more than two weeks now and has not seen one twister.  Last week, Weather Channel Senior Meteorologist Stu Ostro speculated that global warming was the cause. Of course this is the normal, mother nature has a perverse sense of humor - projects to study east coast storms had no east coast storms that winter, just one passing front. You just knew that VORTEX2 would lead to a lack of storms to study. We need them to schedule a massive study of hurricanes and we will surely have a dud season.

As to Stu’s reasoning, a big ridge would lead to heat and dryness across the central states. Instead the region has been hit hard with a steady stream of progressive troughs with heavy rainfall and very little warmth (fuel for storms). These storms brought more toirnadoes in April 2009 than in April 2008.

image
See larger image here.

The real issue may instead be that the El Nino like conditions that came in late May last year shutting off severe weather activity in June 2008 came on a month earlier this year and severe weather activity has diminished in May.

Posted on 05/26 at 02:57 AM
(67) TrackbacksPermalink


Sunday, May 24, 2009
Why NOAA Has Become the Outlier

By Joseph D’Aleo

We have noted in the last year that NOAA has often become the warmest of the 5 major data sets in their monthly globval anomalies. Thanks to Bob Tisdale, we have a better idea why that is the case. Bob occasionally cross posts on Watts Up With That and gets a mention on Climate Audit, for newbies, these three sites should be among your favorite bookmarks.

Recent Differences Between GISS and NCDC SST Anomaly Data And A Look At The Multiple NCDC SST Datasets
By Bob Tisdale

In the recent WUWT post Something hinky this way comes: NCDC data starts diverging from GISS, the differences between GISS and NCDC global temperature anomaly data was discussed. I commented that the GISS and NCDC global surface temperature anomaly data relied on two different SST datasets.

NCDC has their own SST anomaly dataset for their global surface temperature product, and they calculate anomalies against the base years of 1901 to 2000. GISS has used the NCDC OI.v2 SST anomaly data since December 1981, and before that they had used the Hadley Centre’s HADSST data. GISS then splices the two datasets together. This post does not discuss the HADSST data, but delves into the differences between the multiple NCDC SST anomaly datasets, one of which is used by GISS.

GRAPHS OF GLOBAL OI.v2 (USED BY GISS) and “NCDC Ocean” SST ANOMALY DATA

I have not been able to find GISS SST anomaly data as a separate dataset, so for a short-term comparison, I’ll use their source, the OI.v2 SST anomaly data available through the NOAA NOMADS system. Unfortunately, the OI.v2 SST data uses a third climatology for their anomalies (with base years of 1971-2000), but don’t let that concern you. It just makes for an unusual comparative graph.

Figure 1 is a short-term comparison (November 1981 to April 2009) of the OI.v2 Global SST anomaly data (used by GISS) and the NCDC’s “Global Ocean Temperature”.

image
See larger image here.

The two datasets appear to track one another, and the obvious difference, the shift in the data, is a result of the different base years. But if we subtract the OI.v2 SST data from the NCDC “Global Ocean” SST anomaly data, we can see that one dataset rose more than the other since November 1981. Refer to Figure 2. The NCDC “Global Ocean” SST anomaly dataset rose at a greater rate than the OI.v2 SST anomaly data that’s used by GISS. This would bias the NCDC global surface temperature upward over this time span, or bias the GISS data down, depending on your point of view.

image
See larger image here.

Reynolds, Smith, and Liu gave the reasons for removing the satellite data from ERSST.v3 data as, “In the ERSST version 3 on this web page WE HAVE REMOVED SATELLITE DATA from ERSST and the merged product. The addition of satellite data caused problems for many of our users. Although, the satellite data were corrected with respect to the in situ data as described in reprint, there was a residual cold bias that remained. The bias was strongest in the middle and high latitude Southern Hemisphere where in situ data are sparse. THE RESIDAL BIAS LED TO A MODEST DECREASE IN THE GLOBAL WARMING TREND AND MODIFIED GLOBAL ANNUAL TEMPERATURE RANKINGS.”

image
See larger image here.

Hmmm. It looks as though, if you’re a SST data producer, downward biases are bad, but increases in trend with each update are good. See much more
detail and get data links here.

Icecap Note: NOAA and the other ground based data centers would have more credibility if one of the changes resulting in a reduction of the warming trend and not an exaggeration. Anthony Watts has clearly shown increasingly bad siting can lead to warm bias here. NOAA’s allowing 2/3rds of the world’s stations to dropout in 1990 led to more urban bias and thus warming, and NOAA has removed the US Urban Heat Island (UHI) adjustment and performs no UHI adjustment on global data. This is despite the facts that NCDC’s own Director Tom Karl in Kark et al (1988 J Climate) in Urbanization: its detection and effect in the United States climate record, showed the importance of urban adjustment and the Hadley Centre’s Phil Jones (2008) in Jones et al. ”Urbanization effects in large-scale temperature records, with an emphasis on China”, showed UHI’s contamination of data in China.

Removal of the UHI for the US resulted in a warming relative to GISS with UHI in the United States of 0.75F since 1940. Now having gotten all the warmth possible out of the land temperatures, they turn to the oceans, so promising as they cover 70% of the earth’s surface.

image
See larger image here.

Posted on 05/24 at 03:11 AM
(269) TrackbacksPermalink


NIPCC report states there is no anthropogenic global warming

By Dianna Cotter, Portland Civil Rights Examiner

In May 2006 we were told, with the most alarming language possible, that humans were causing the planet to warm with their use of Carbon based fuels. We must follow all of the recommendations of the IPCC report coming out in 2007 or humanity’s existence on earth would end. We were told that we must do something now, in order to prevent the deaths of untold millions, and the loss of humanity’s very ability to survive on earth. We must start taxing “Global warming Gasses” in order to stop their production, no matter the human cost! It’s a complete lie.

The IPCC - the International Panel on Climate Change is a committee with an agenda, and one it was pre-programmed to execute. It was designed from the outset, from its very beginnings to come up with a specific result. This is in direct contradiction to every principle of real science, the aim of which is to describe reality as it is, not reality as one wishes it to be.

To illustrate this, in 1995 the IPCC completely ignored Satellite evidence, very clear evidence, that there was no warming according to a report from the Heartland Institute Released in 2008. The full pdf is titled Nature, Not Human Activity, Rules the Climate . The institute claims, and is backed by anecdotal reports that the IPCC changed portions of the text of its report to make it appear that humans were the cause of the warming they artificially detected, or in other words, manufactured. Also attributed to the Heartland Institute’s report, the IPCC ignored further data confirming the lack of warming - if not showing a slight cooling - that became available after the May 2006 deadline.

The report the column speaks of is the NIPCC. The Nongovernmental International Panel on Climate Change. A completely independent examination of all the evidence available in published peer-reviewed literature. They included every single piece or evidence available to them, positive and negative, they did not select research that proved their predetermined result as the IPCC did. They went through everything. Their conclusion, there is no human caused global warming.

“The 1995 IPCC report was notorious for the significant alterations made to the text after it was approved by the scientists - in order to convey the impression of a human influence. The 2001 IPCC report claimed the twentieth century showed ‘unusual warming’ based on the now-discredited hockey-stick graph. The latest IPCC report, published in 2007, completely devaluates the climate contributions from changes in solar activity, which are likely to dominate any human influence”.

For anyone who has any sort of intellectual honesty at all, the report from the Heartland Institute is a MUST read. Dr. Fred S. Singer who conceived and directed the NIPCC project is an honored and decorated scientist. His Raison d’etre is truth, not political machinations, or power. He is deeply concerned by the absolute mis-use of science in what has now become a multi-billion dollar business, that is based on what amounts to a hoax.

“The Fourth Assessment Report of the IPCC [IPCC-AR4 200 “When new errors and outright falsehoods were observed in the initial drafts of AR4 (The Fourth Assessment Report of the IPCC [IPCC-AR4 2007]), SEPP [Science and Environmental Policy Project] set up a ‘Team B’ to produce an independent evaluation of the available scientific evidence. While the initial organization took place at a meeting in Milan in 2003, ‘Team B’ was activated only after the AR4 SPM appeared in February 2007. It changed its name to NIPCC and organized an international climate workshop in Vienna in April 2007.”

This group of scientists, who have looked at all available evidence, from satellite to very unreliable ground temperature readings to historical records, concludes correctly; if Human governments are going to be institution policies that will drastically affect the lives of its citizens, and the Waxman-Markey Cap and trade energy bill is without a doubt just that, then it must be utterly certain of the science that backs the supporting reasons for the existence of the legislation.

This report states unequivocally that there is zero evidence of man-made global warming. We are about to kill a Nation, literally for hot air that does not exist. Considering the likely ramifications of Waxman-Markey Cap and trade energy bill, indeed, we must have the fortitude to re-examine honestly all of the information available to us. For those of you who believe whole heartedly in Man-made Global warming, then these reports are a must read for you. They are peer reviewed documents based in fact. If you are supporting legislation that will wreck such economic devastation not only on your fellow Americans but also on yourself, it should be read so that you can counter those who do not believe in global warming caused by man. Because you will have some explaining to do. Be cautioned however, upon reading them, you will find yourself starting to realize that you have been had.  Human caused Legislation like the Waxman-Markey Cap and trade energy bill will do more harm to more people in a year, than human caused Global Warming will in the next century. It is nothing short of a massive money stealing scheme made by our Governement to pay for the whorish excesses of the last 6 months. Now that is an inconvenient truth. Read more here. H/T Climate Depot.

Posted on 05/24 at 01:20 AM
(79) TrackbacksPermalink


Saturday, May 23, 2009
In the UK the media is getting very excited about wind power

By Ann McElhinney & Phelim McAleer

In the UK the media are getting very excited about wind power.

But as we discussed before environmental journalists seem to be little more than nodding dogs who endlessly believe and cheerlead everything they are told by activist groups.

Bringing truth to power seems to disappear when the power is a well funded environmental group. And so it is with wind power. Recently the BBC and Channel 4 have carried big excited stories about a planned new wind farm in UK. However neither felt it important to EVEN mention the little issue of “back up”. Both reports made much of the wind farm being able to power 180,000 homes which they stated could produce enough energy to power Glasgow.

But the nodding dogs are being dishonest and disingenuous. Wind will never power towns or cities on their own and giving that impression is both dangerous and misleading. We still need fossil fuels to keep us safe and the lights on as reported in this Reuters story.  “Britain, which aims to install about 30 gigawatts (GW) of wind turbines by 2020, will need to build almost as much backup power generation for calm weather periods, an executive from Scottish Power said on Wednesday,” Reuters reported.

“Thirty gigawatts of wind maybe requires 25 GW of backup,” said Rupert Steele, regulation director at the Scottish arm of Spain’s Iberdrola, one of the world’s largest wind farm operators. “The problem is that if you’ve got a high-pressured area, you may have quite a large area where there’s no wind at all ... That happens also offshore,” he told Reuters.”

image

Environmentalists know this. Just ask them to send their child to a hospital whose system is fully dependent on wind power and watch them run in the other direction. They know that wind is erratic and will never supply the electricity that keeps our towns, cities factories, schools and hospitals running and all of us safe and healthy. But they don’t tell us this when they triumphantly pronounce that we can power the world with solar and wind. What is so much more disappointing is that so called ‘impartial’ reporters don’t tell us either.

Shame on you Channel 4 and BBC for not reporting this truth to your viewers.

See earlier story “The Answer My Friend is NOT Blowing in the Wind” here.

See Congresswoman Michele Bachmann (R-MN) in this Washington Observer Interview says its so well here.

See more about Ann and Phelim’s important Not Evil, Just Wrong project.

See in this UK register story, about another green favorite boondoggle, the electric car. The Politicians’ obsession with electric cars is a waste of time - and costing British science and research dearly. So says Richard Pike, head of the Royal Chemistry Society, in a hard-hitting contribution to Research Fortnight (pdf).

Posted on 05/23 at 01:01 AM
(61) TrackbacksPermalink


Thursday, May 21, 2009
Follow the Science on Yucca

New York Times Editorial

The administration’s budget for the Energy Department raises a disturbing question. Is President Obama, who has pledged to restore science to its rightful place in decision making, now prepared to curtail the scientific analyses needed to determine whether a proposed nuclear waste repository at Yucca Mountain in Nevada would be safe to build?

image

It is no secret that the president and the Senate majority leader, Harry Reid, who hails from Nevada, want to close down the Yucca Mountain project, which excites intense opposition in the state. The administration has proposed a budget for fiscal year 2010 that would eliminate all money for further development of the site, and Mr. Reid has pronounced the project dead.

But the administration at least claimed that it would supply enough money for the Energy Department to complete the process of seeking a license from the Nuclear Regulatory Commission, if only to gain useful knowledge about nuclear waste disposal. Unfortunately, the budget released this month looks as if it will fall well short of the amount needed.

Money for the Yucca Mountain project, nearly all of which is used to support the licensing application, would fall from $288.4 million in 2009, the current year, to $196.8 million in 2010, a precipitous drop. And the agency intends to rely heavily on its own staff personnel rather than on more costly outside consultants from the national laboratories or private contractors. There is great danger that the department will lack the expertise needed to answer tough technical questions that emerge during the regulatory commission’s reviews.

These ramp-downs are occurring at the worst time. The regulatory commission is just beginning its licensing process, which is scheduled to take three to four years, and its relevant boards have ruled that at least eight intervenors can raise some 300 issues for technical challenges, an unusually high number. The cutbacks increase the odds that the agency will stumble in trying to justify a license - or that the hearings and evaluations won’t be completed within statutory deadlines.

Meanwhile, the administration, Congressional leaders and the nuclear industry are calling for a blue-ribbon panel to study alternative ways to dispose of nuclear waste. Surely it would be useful for any such panel to know whether the Yucca Mountain project was sound or flawed.

Before approving this truncated budget, Congress needs to ensure that it contains enough money to sustain a genuine licensing effort. We have no idea whether Yucca Mountain would be a suitable burial ground for nuclear wastes. But after the government has labored for more than two decades and spent almost $10 billion to get the site ready for licensing hearings, it would be foolish not to complete the process with a good-faith evaluation. Are Mr. Obama and Mr. Reid afraid of what the science might tell them? Read editorial here.

Posted on 05/21 at 08:28 PM
(83) TrackbacksPermalink


Tuesday, May 19, 2009
Comments On The EPA “Proposed Endangerment And Cause Or Contribute Findings”

By Roger Pielke Sr., Climate Science

I have generally supported most EPA actions which have been designed to support environmental improvement. These regulations have resulted in much cleaner water and air quality over the past several decades; e.g. see National Research Council, 2003: Managing carbon monoxide pollution in meteorological and topographical problem areas. The National Academies Press, Washington, DC, 196 pp.

However, the EPA Endangerment Findings for CO2 as a climate forcing falls far outside of the boundary of the type of regulations that this agency should be seeking. The EPA on April 17, 2009 released this finding in ”Proposed Endangerment and Cause or Contribute Findings for Greenhouse Gases under the Clean Air Act”.  This report is a clearly biased presentation of the science which continues to use the same reports (IPCC and CCSP) to promote a particular political viewpoint on climate (and energy policy).

“After a thorough examination of the scientific evidence on the causes and impacts of current and future climate change, as well as other effects of greenhouse gases, the Administrator concludes that the science compellingly supports a positive endangerment finding for both public health and welfare. In her decision, the Administrator relied heavily upon the major findings and conclusions from recent assessments of the U.S. Climate Change Science Program and the Intergovernmental Panel on Climate Change.” “The Administrator is proposing this endangerment finding after considering both observed and projected future effects of climate change, key uncertainties, and the full range of risks and impacts to public health and welfare occurring within the United States. The scientific evidence concerning risks and impacts occurring outside the United States, including risks and impacts that can affect people in the United States, provides further support for this proposed endangerment finding.”

What these statements tell us is that their finding includes results from multi-decadal climate predictions, which have never shown regional predictive skill, including any ability to predict past major weather events such as droughts! Nevertheless, they make claims with respect to the climate risks as if these are certain, despite the absence of skill in predicting them for the 20th century. They claim that

“The effects of climate change observed to date and projected to occur in the future include, but are not limited to, more frequent and intense heat waves, more severe wildfires, degraded air quality, more heavy downpours and flooding, increased drought, greater sea level rise, more intense storms, harm to water resources, harm to agriculture, and harm to wildlife and ecosystems. The Administrator considers these impacts to be effects on public health and welfare within the meaning of the Clean Air Act.”

In the document “Frequently Asked Questions on the Proposed Endangerment and Cause or Contribute Findings for Greenhouse Gases”, they have the Q&A. On what science was the proposed Endangerment Finding based?

“The Administrator relied heavily on existing, peer-reviewed scientific literature. In particular, she relied on reports and conclusions from the U.S. Climate Change Science Program, the National Research Council, and the Intergovernmental Panel on Climate Change because they represent the current state of knowledge on climate change science, vulnerabilities, and impacts. These studies are authored by leading scientific experts and underwent multiple layers of peer review, including, in many cases, review and acceptance by government agencies.”

As documented in Pielke Sr., Roger A., 2008: A Broader View of the Role of Humans in the Climate System is Required In the Assessment of Costs and Benefits of Effective Climate Policy. Written Testimony for the Subcommittee on Energy and Air Quality of the Committee on Energy and Commerce Hearing “Climate Change: Costs of Inaction” - Honorable Rick Boucher, Chairman. June 26, 2008, Washington, DC., 52 pp.; Pielke Sr., Roger A., 2005: Public Comment on CCSP Report ”Temperature Trends in the Lower Atmosphere: Steps for Understanding and Reconciling Differences”. 88 pp including appendices; and Protecting The IPCC Turf - There Are No Independent Climate Assessments Of The IPCC WG1 Report Funded And Sanctioned By The NSF, NASA Or The NRC.

The conflict of interest (with most of the same individuals leading the reports) is easy to see; i.e. the Federal (non-EPA) expert reviewers for the EPA Endangerment findings (see ”Technical Support Document for the Proposed Findings” are

Virginia Burkett, USGS; Phil DeCola; NASA (on detail to OSTP); William Emanuel, NASA; Anne Grambsch, EPA; Jerry Hatfield, USDA; Anthony Janetos; DOE Pacific Northwest National Laboratory; Linda Joyce, USDA Forest Service; Thomas Karl, NOAA; Michael McGeehin, CDC; Gavin Schmidt, NASA; Susan Solomon, NOAA; Thomas Wilbanks, DOE Oak Ridge National Laboratory.

In conclusion, the EPA Endangerment findings is the culmination of a several year effort for a small group of climate scientists and others to use their positions as lead authors on the IPCC, CCSP and NRC reports to promote a political agenda. Now that their efforts have reached the federal policy decision level, Climate Science urges that there be an independent commission of climate scientists who can evaluate the assement process that led to the EPA findings as well as the climate science upon which it is constructed. Read full comments here.

Posted on 05/19 at 03:53 PM
(293) TrackbacksPermalink


Saturday, May 16, 2009
Is The U.S. Temperature Record Reliable?

By Mish Michaels, WBZ-TV, Boston

Billions of dollars are being spent on “fixing” the climate issue, but how much is invested in data collected to identify the problem? There has clearly been a warming trend over the last century, but by how many degrees? A new report published by meteorologist Anthony Watts draws into question the reliability of temperature data collected at over 1,200 sites across the country.

These sites represent the U.S. Historical Climate Network (USHCN). Some of the surface stations have been collecting temperature data for over 100 years. Working with hundreds of volunteers, Watts has thoroughly documented the environment surrounding some 80 percent of surface stations from coast to coast—an effort never done before.

THE REPORT’S FINDINGS:

PDF: Are Surface Stations Reliable?

According to the U.S. government, a surface station must be at least 100 feet away from a heat source/sink to be considered reliable. Using the goverment’s own standards for properly locating temperature sensors, Watts graded each site on a scale from 1 to 5. A grade of 1 or 2 indicates reliable placement. A grade of 3 to 5 can result in temperature errors of several degrees, according to the governments own studies.

To date, the investigation has noted that only 3 percent of the stations surveyed were grade 1. Roughly 8 percent were grade 2, meaning only 11 percent of the stations are located in a manner that results in reliable temperature data. Some 20 percent of stations were grade 3, the majority were grade 4 (58 percent), and 11 percent were graded as 5.

image

Locally, the Blue Hill Observatory (shown above and enlarged here) got a grade of 3. A site in Bedford recently closed down was rated as a 4. There was a site in Chestnut Hill but it was closed down long ago. A notation on the site indicated the “temperature data too unreliable"-- yet the decades it did collect information is part of the climate record. There are no sites in our state considered to be 1 or 2. The audit revealed temperature recording stations next to exhaust fans, air condition units, surrounded by asphault roads, on rooftops, near sidewalks, and close to buildings. All of these radiate and absorb heat. Remarkably, 68 stations were located at wastewater treatment plants.

View the audit of surface stations in Massachusetts. View the audit of surface stations by state. Amherst, MA shown.

image

THAT IS NOT ALL; OTHER CONCERNS NOTED

1) To collect and communicate the data, the goverment relies on citizen volunteers from private folks to prisoners in jail. It is a triumph of service yet these efforts result in missing data—roughly 30 percent.
2) Instrument shelters were painted with white wash paint since the late 1800s. Due to concerns over the lead based paint, a few decades ago laytex replaced the white wash. Studies have shown that the laytex runs warmer.
3) In the 1980s, new instrument towers were introduced called MMTS. Because the output screen needs to be placed indoors, the unit must be cabled to a building, raising many siting issues due to this limitation.

This week 150,000 copies of the surface stations report landed on the desk of politicians.

The National Weather Service (NWS) and the National Climatic Data Center (NCDC) are well aware of the issues that exist and have been for some time. In response they have established a new automated temperature network at 114 sites called the CRN--Climate Reference Network. The sites will take measurements every 30 seconds, averaged out every 5 minutes. Temperatures are then transmitted hourly.Originally there were to be 250 sites with two sites in close proximity for redundancy. Budget cuts scaled things back. These sites are considered to ensure the highest quality of temperature data and are meant to tease out national climate trends. In New England, there are two CRN sites in Maine, one site with redundancy (two in close proximity) in Durham, New Hampshire and in Kingston, RI. There are no CRN sites in Massachusetts. The network was only recently established, but so far the data collected is said to match quite well with the temperature data from the USHCN.

The government also plans to establish a much larger network called the HCNM---basically the M stands for “modernized.” This other network will be 1000 sites strong across the country and will have the density to capture regional climate trends. The surface stations will also be automated and will be located on public lands to prevent any urban encroachment or the need to move the stations. So far none of the instrumentation has been installed, only prototypes are being tested in the field at this time. Due to budget constraints, this new network will take years, maybe even more than a decade to install. The Blue Hill Weather Observatory is being inspected as a potential site for addition to this new network. Read this full post here. Kudos to Mish Michaels and WBZ for their willingness to present this critical issue at this very critical time when very important decisions are being made in Washington. Got it mostly right. They were not aware NOAA removed the urban adjustment in 2007.

Posted on 05/16 at 01:49 AM
(449) TrackbacksPermalink


Friday, May 15, 2009
Clarity on Cap-and-Trade

By Donn Dears

Describes workings of Cap & Trade, the legislation designed to force the U.S. to cut its CO2 emissons. Only 24% of the public even know it has something to do with energy and the environment. Hopefully this video and the follow up coming soon will make us all smarter about what it is and what its impact will be.

Full size youtube here.

Brief Overview of Donn Dears Experience with Energy Issues
Donn Dears began his career at General Electric testing large steam turbines and generators used by utilities to generate electricity; followed next, by manufacturing and marketing assignments at the Transformer Division. He next led an organization of a few thousand people servicing these and other GE products in the United States. He then established facilities around the world to service power generation and transmission equipment. At nearly every step, Donn was involved with the work done at customer locations; at steel mills, electric utilities, refineries, oil drilling and production facilities and open pit and underground mining operations. At every opportunity, he learned of the needs of these industries.

Donn has toured Saudi Arabia including a close-up inspection of the eastern province with its oil producing and shipping facilities. He has investigated many of the other oil producing countries in the Mid East and Northern Europe, as well as examining iron-ore mining locations and major shipping centers in Europe and Asia. All told, Donn has visited over 50 countries and has knowledge of their need for the technologies that can improve their well being and their use of equipment manufactured in the United States.

Following his retirement as a senior GE Company executive, he continued to study and write about energy issues. He currently is president of TSAugust, a 501 (C) 3 think tank comprised entirely of volunteers. He writes for http://www.tsaugust.org and has written two books as well as various papers and articles. He also speaks about energy issues.

Donn is a graduate of the United States Merchant Marine Academy and served on active duty in the U.S. Navy

Posted on 05/15 at 03:48 AM
(0) TrackbacksPermalink


Tuesday, May 12, 2009
Cities to Sizzle as Islands of Heat

By Jonathan Leake, Environment Editor Timesonline

LONDON and other cities could see summer temperatures rise to more than 10C above those in the surrounding countryside, according to Met Office research being used to help devise the first official climate change map of Britain. Scientists have been studying a phenomenon known as the urban heat island effect, in which cities become significantly hotter than the areas around them because of the heat they generate themselves.

Big cities such as London, Birmingham, Manchester and Glasgow already reach temperatures 2C - 3C above their surroundings in the summer. Scientists fear that difference could grow four to fivefold as hotter weather combines with soaring energy use and population growth, making such temperature gaps more frequent and more extreme.

The research is linked to a wider project aimed at helping scientists predict the impact rising temperatures will have on different parts of the country. The full results will be released next month by Hilary Benn, the environment secretary. Vicky Pope, the head of climate change advice at the Met Office, said: “As the climate gets warmer, sweltering summer temperatures will combine with rising energy use, the heat-retaining properties of buildings, and the sheer volume of people, to push temperatures higher and higher. It may sometimes make life in the metropolis intolerable. Imagine the scorching conditions that commuters will face on London’s Tube network.”

The warning follows the disclosure by the Intergovernmental Panel on Climate Change that global temperatures have risen by almost 1C since preindustrial times. The panel predicts global temperatures will have risen by 2C by 2050 with total warming of up to 5-6C possible by 2100. Such findings are now widely accepted but questions remain, especially regarding the impact on cities, where more than half the world’s population live. New York - hotter in summer than British cities - is regularly 7C-8C hotter than nearby rural areas.

In Britain, 90% of the population lives in urban or suburban areas so the impact on people is potentially huge. The research is based partly on data from heatwaves, such as the one in 2003, and on computer projections. It also looked at cities such as Athens and Beirut which suffer from the urban heat phenomenon. The August 2003 heatwave saw England’s daytime temperatures top 30C for 10 days and exceed 35C in many places. The same heatwave saw temperatures in the upper 30Cs in the centres of cities such as London, Birmingham and Manchester. This was often 6C-7C above those in rural areas.

Researchers fear central city temperatures may exceed 40C as the century progresses. “The high temperatures of 2003 were extraordinary but may become common by 2050 and even be seen as relatively cool by 2100,” said Pope. One of the factors that made London so hot was its inability to cool down. At night during the heatwave, the city centre was sometimes 9C warmer than its surrounding green belt.

This is because rural and suburban areas lose heat at night but in cities the materials used for hard surfaces store more solar energy and lose it more slowly. This effect is amplified by the heat from lights, electrical equipment and cars. Also, as cities get warmer, they consume more power trying to stay cool, because of airconditioners and fridges working harder. Richard Betts, head of climate impacts at the Met Office, who oversaw the research, said Tokyo showed what British cities might face. Its tall, densely packed buildings and high energy use mean the Japanese capital is often 10C hotter than the surrounding countryside. “We must change how we plan cities, to maximise green spaces and create structures that dissipate heat,” said Betts.

Urban heat islands have a serious impact on health. In 2003 there were 2,091 more deaths than normal between August 4 and 13 in Britain, most of them among elderly people in southeast England. For people aged over 75 there was a 33% increase in mortality.

The UKMO is promoting and the author is endorsing the IPCC forecasts which are failing miserably. They are ironically now “studying” the Urban Heat Island (UHI) effect that was said to not be important to global temperatures. The Hadley Center (like NOAA) makes no adjustment for UHI based on Jones et. al. (1990) and later Parker (2004) who claimed that the non-climatic bias due to urbanization is less than one-tenth of the global trend. Recently Jones coauthored a paper that showed significant UHI contamination in China but he claimed it did not apply globally and now the UKMO is using it to pump up the scare tactics for the UK. Read more on the UHI here.

image
In the United states, the number of state record highs has been lower this decade than any decade since the 1800s.

Posted on 05/12 at 10:13 PM
(0) TrackbacksPermalink


SPPI Monthly CO2 Report - April 2009

SPPI’s authoritative Monthly CO2 Report for April 2009 reveals that:

(1) Unpredicted ocean cooling over the past five years disproves the theory that manmade “global warming” has a significant effect on the Earth’s temperature. None of the IPCC’s computer models had predicted ocean cooling. All had predicted ocean warming. They were wrong.

image
See larger image here

(2) Rapid surface atmospheric cooling, especially in the Northern Hemisphere, has now continued for seven and a half years.

image
See larger image here

(3) The IPCC assumes CO2 concentration will reach 836 ppmv by 2100. However, for seven years, CO2 concentration has headed straight for only 575 ppmv by 2100. This alone halves all of the IPCC’s temperature projections.

(4) Since 1980 temperature has risen at only 2.5 F (1.5 C) per century, not the 7 F (3.9 C) the IPCC imagines. For 600 million years there has been no correlation between CO2 concentration and the Earth’s temperature.

image
See larger image here

(5) Sea level rose just 8 inches in the 20th century and has been rising at just 1 foot/century since 1993. Though James Hansen of NASA says sea level will rise 246 feet, sea level has scarcely risen since the beginning of 2006.

image
See larger image here

(6) Sea ice extent in the Arctic is above the 30-year average, and has set a nine-year record high. In the Antarctic, sea ice extent reached a record high in 2007, and is now the third-highest in 30 years. Global sea ice extent shows little trend for 30 years.

(7) The Accumulated Cyclone Energy Index is a 2-year running monthly sum of activity in all hurricanes, typhoons and tropical cyclones. It shows that there is now less severe tropical-storm activity than at any time in 30 years.

(8) Solar activity is at a 100-year record low. We may be facing a 70-year Maunder Minimum - extreme cooling.

(9) Science Focus this month studies “data revisionism” - how scientific results are bent to promote false alarmism.

image
See larger image here of GISS raw versus adjusted data for Santa Rosa, one of many such examples.

Finally, check out our monthly selection of scientific papers, and meet the Top Ten “Global Warming” skeptics in the full report here.

Be sure also to see Lord Monckton respond back to Gavin Schmidt’s attacks in this SPPI post here. He begins “Yet again Gavin Schmidt, who writes the political RealClimate blog for “global-warming” alarmists that is promoted by two of the authors of the now-discredited “hockey-stick” graph falsely abolishing the medieval warm period, has inaccurately criticized me in his blog. This paper sets the facts straight. On 2 May 2009, under the title Monckton’s deliberate manipulation, Gavin Schmidt, who is paid by NASA but may not be as scientific or as impartial as a government servant ought to be, posted a blog entry that we shall publish in full, paragraph by paragraph, with our commentary in bold face.”

Posted on 05/12 at 08:06 PM
(0) TrackbacksPermalink


Page 65 of 97 pages « First  <  63 64 65 66 67 >  Last »
Blogroll

Climate Debate Daily

Climate Change Fraud

Science Bits

Climate Skeptic

Science and Public Policy Institute

Gary Sharp’s It’s All About Time

CO2web

Global Warming Hoax

Accuweather Global Warming

Tallbloke

Landsurface.org, The Niyogi Lab at Purdue

CO2 Sceptics

Climate Cycle Changes

Hall of Record

The Inhofe EPW Press Blog

Scientific Alliance

APPINYS Global Warming

CO2 Science

Carbon Folly

TWTW Newsletters

The Weather Wiz

Wisconsin Energy Cooperative

Climate Depot

Dr. Roy Spencer

Gore Lied

James Spann’s Blog

Musings of the Chiefio

Science and Environmental Policy Project

Middlebury Community Network on The Great Global Warming Hoax

The Resilient Earth

Blue Hill Observatory, Milton MA

Blue Crab Boulevard

Global Warming Skeptics

Craig James’ Blog

Climate Debate Daily

Anthony Watts Surface Station Photographs

Digging in the Clay

Tom Skilling’s Blog

Global Warming Scare

Ice Age Now

Analysis Online

Dr. Roy Spencer

Watts Up with That?

Cornwall Alliance

Committee for a Constructive Tomorrow (CFACT)

Climate Debate Daily

John Daly’s What the Stations Say

Earth Changes

Joanne Nova- The Skeptic’s Handbook

Energy Tribune

I Love My Carbon Dioxide

Climate Science: Roger Pielke Sr. Research Group Weblog

Greenie Watch

Right Side News

Web Commentary

Global Warming Hoax

The Reference Frame - Lubos Motl’s weblog

Roy Spencer’s Nature’s Thermostat

Warmal Globing

The Heartland Institute

Bill Meck’s Blog

Raptor Education Foundation

John McLean’s Global Warming Issues

The Week That Was by Fred Singer

Where is Global Warming (Bruce Hall Collection)

Weatherbell Analytics

Dr. Dewpoint on Intellicast

COAPS Climate Study US

John Coleman’s Corner

Raptor Education Foundation

Climate Research News

The Cornwall Alliance

Warwick Hughes

Intellicast Dr. Dewpoint

AMSU Global Daily Temps

Art Horn’s “The Art of Weather”

Tom Nelson Blogroll

Finland Lustia Dendrochronology Project

Demand Debate

Climate Audit

Vaclav Klaus, Czech Republic President

Marshall Institute Climate Change

Reid Bryson’s Archaeoclimatology

The Climate Scam

Prometheus

World Climate Report

Climate Resistance

Climate Police

Tropical Cyclone Blog of Ryan Maue COAPS

Ross McKitrick Google Home Page

Carbonated Climate

Omniclimate

Finland Lustia Dendrochronology Project

Redneck USA

Bald-Faced Truth

The New Zealand Climate Science Coalition

Junk Science

MPU Blog

Bob Carter’s Wesbite

Powerlineblog